Minimum distance histograms with universal performance guarantees
نویسندگان
چکیده
منابع مشابه
Optimal Histograms with Quality Guarantees
Histograms are commonly used to capture attribute value distribution statistics for query optimizers. More recently, histograms have also been considered as a way to produce quick approximate answers to decision support queries. This widespread interest in histograms motivates the problem of computing his-tograms that are good under a given error metric. In particular, we are interested in an e...
متن کاملSmooth Interpolating Histograms with Error Guarantees
Accurate selectivity estimations are essential for query optimization decisions where they are typically derived from various kinds of histograms which condense value distributions into compact representations. The estimation accuracy of existing approaches typically varies across the domain, with some estimations being very accurate and some quite inaccurate. This is in particular unfortunate ...
متن کاملGraphs with distance guarantees
One goal in network design is the construction of sparse networks that guarantee short distances with respect to some given distance requirements. By this, it can be guaranteed, for example, that delays that are incurred by link faults are bounded. An appropriate graph-theoretic model for this is the concept of k–spanners: Given a graph G, a k–spanner of G is a spanning subgraph S, such that th...
متن کاملStreaming self-scaling histograms with stability and optimality guarantees
Histograms provide discrete efficient representation of continuous data and can be constrained to fit various purposes. We provide the first offline sub-quadratic algorithm which applies to any sub-additive histogram cost functions. Sub-additive cost functions includes in particular all the `p norms. This algorithm outputs a histogram with at most 2m buckets and cost at most (1 + ε) the optimal...
متن کاملNon-convex boosting with minimum margin guarantees
Many classification algorithms achieve poor generalization accuracy on “noisy” data sets. We introduce a new non-convex boosting algorithm BrownBoost-δ, a noiseresistant booster, that is able to significantly increase accuracy on a set of noisy classification problems. Our algorithm consistently outperforms the original BrownBoost algorithm, AdaBoost, and LogitBoost on simulated and real data. ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Japanese Journal of Statistics and Data Science
سال: 2019
ISSN: 2520-8756,2520-8764
DOI: 10.1007/s42081-019-00054-y